# Long context processing

A.X 4.0 Light Gguf
Apache-2.0
A.X 4.0 Light is a lightweight large language model developed by SKT AI Model Lab, built on Qwen2.5 and optimized for Korean understanding and enterprise deployment.
Large Language Model Transformers Supports Multiple Languages
A
mykor
535
2
Longwriter Zero 32B I1 GGUF
Apache-2.0
The LongWriter-Zero-32B quantized model is based on the THU-KEG/LongWriter-Zero-32B base model, supports both Chinese and English, and is suitable for long context scenarios such as reinforcement learning and writing.
Large Language Model Transformers Supports Multiple Languages
L
mradermacher
135
1
Longwriter Zero 32B GGUF
Apache-2.0
The LongWriter-Zero-32B Quantized Model is a multilingual model that undergoes static quantization based on the original model. It is suitable for long context scenarios such as reinforcement learning and writing.
Large Language Model Transformers Supports Multiple Languages
L
mradermacher
204
1
Jan Nano Gguf
Apache-2.0
Jan Nano is a small language model fine - tuned based on the Qwen3 architecture, designed for local and embedded environments, combining efficiency and long - context processing capabilities.
Large Language Model
J
Menlo
2,703
38
Deepseek R1 0528 GPTQ Int4 Int8Mix Compact
MIT
The GPTQ quantized version of the DeepSeek-R1-0528 model, using a quantization scheme of Int4 + selective Int8, which reduces the file size while ensuring the generation quality.
Large Language Model Transformers
D
QuantTrio
258
1
Zlatorog 12B Instruct Beta
Zlatorog-12B-Instruct-Beta is a large language model fine-tuned based on Mistral Nemo 12B, designed specifically for Slovenian language-related research, supporting instruction following, reasoning, and tool-enhanced generation tasks.
Large Language Model Transformers Supports Multiple Languages
Z
zID4si
123
1
Qwen3
Apache-2.0
Qwen3 is a series of advanced language models developed by the Qwen Team, excelling in advanced reasoning, dialogue, instruction following, and agent applications.
Large Language Model
Q
cortexso
8,089
2
Bamba 9B V2
Apache-2.0
Bamba-9B-v2 is a decoder-only language model built on the Mamba-2 architecture, focusing on text generation tasks and outperforming Llama 3.1 8B in performance.
Large Language Model Transformers
B
ibm-ai-platform
3,634
15
Llama Krikri 8B Instruct GGUF
A Greek instruction-tuned large language model based on Llama-3.1-8B, enhancing Greek language capabilities and supporting multilingual tasks
Large Language Model Transformers
L
ilsp
257
11
Llama Krikri 8B Base
Llama-Krikri-8B-Base is a large Greek foundational language model built upon Llama-3.1-8B, extending Greek language capabilities through continued pretraining while maintaining English proficiency.
Large Language Model Transformers Supports Multiple Languages
L
ilsp
104
12
Qwen2.5 14B DeepSeek R1 1M
Apache-2.0
A multi-purpose large language model that combines the advantages of reasoning models and long context models
Large Language Model Transformers
Q
mkurman
2,195
52
Deepseek R1 Bf16
MIT
DeepSeek-R1 is the first-generation inference model, which performs excellently in mathematics, code, and reasoning tasks, and its performance is comparable to that of OpenAI-o1.
Large Language Model Transformers
D
opensourcerelease
1,486
16
Lucie 7B
Apache-2.0
Lucie-7B is a multilingual causal language model with 7 billion parameters jointly built by LINAGORA and OpenLLM-France. It is based on the Llama-3 architecture and pre-trained on 3 trillion tokens of multilingual data.
Large Language Model TensorBoard Supports Multiple Languages
L
OpenLLM-France
1,262
20
Yi 1.5 9B Chat 16K Abliterated
Apache-2.0
Based on the ablated version of the large language model, with better performance than the original model, supporting Chinese, English, and long context processing.
Large Language Model Transformers Supports Multiple Languages
Y
byroneverson
10.52k
5
AI21 Jamba Large 1.5
Other
AI21 Jamba 1.5 is a series of advanced foundation models with powerful long context processing capabilities and efficient inference speed, suitable for various business scenarios.
Large Language Model Safetensors Supports Multiple Languages
A
ai21labs
2,642
216
Jais Adapted 7b
Apache-2.0
The Jais series are specialized bilingual large language models for Arabic processing, adapted for Arabic through pre-training based on the Llama-2 architecture
Large Language Model Supports Multiple Languages
J
inceptionai
1,726
6
Jais Family 590m
Apache-2.0
The Jais series is a bilingual large language model specifically optimized for Arabic while maintaining strong English capabilities. This 590M parameter version is the smallest model in the series, suitable for lightweight applications.
Large Language Model Supports Multiple Languages
J
inceptionai
86
7
Jais Adapted 70b
Apache-2.0
The Jais series is a set of English-Arabic bilingual large language models, optimized for Arabic while maintaining strong English capabilities, with models ranging from 590 million to 70 billion parameters.
Large Language Model Supports Multiple Languages
J
inceptionai
2,579
21
Jais Adapted 70b Chat
Apache-2.0
The Jais series is a large English-Arabic bilingual language model optimized specifically for Arabic. It undergoes adaptive pre-training based on the Llama-2 architecture and has powerful bilingual processing capabilities.
Large Language Model Supports Multiple Languages
J
inceptionai
5,466
9
Meltemi 7B Instruct V1.5
Apache-2.0
Meltemi 7B Instruct v1.5 is a Greek instruction fine-tuned large language model improved based on Mistral 7B, focusing on Greek natural language processing tasks.
Large Language Model Transformers
M
ilsp
1,237
21
Meltemi 7B V1.5
Apache-2.0
Meltemi is a large-scale Greek basic language model trained by the Athens Center for Research and Innovation, built on Mistral 7B and focused on Greek language processing capabilities.
Large Language Model Transformers Supports Multiple Languages
M
ilsp
106
15
Aya 23 35B
Aya 23 is an instruction-fine-tuned open-weight research version model with highly advanced multilingual capabilities, supporting 23 languages.
Large Language Model Transformers Supports Multiple Languages
A
CohereLabs
3,721
282
Evo 1 8k Base
Apache-2.0
Evo is a biological foundation model capable of long context modeling and design. It uses the StripedHyena architecture and can model sequences at single nucleotide and byte-level resolution.
Molecular Model Transformers
E
togethercomputer
31.09k
9
Yi 34B 200K
Apache-2.0
The Yi series of models are next-generation open-source large language models trained from scratch by 01.AI. They support bilingual (Chinese and English) and perform excellently in language understanding, common-sense reasoning, reading comprehension, etc.
Large Language Model Transformers
Y
01-ai
12.63k
317
Extended Mind Mpt 7b
An extended mind Transformer model adjusted based on the Mosaic ML architecture, supporting external memory library retrieval and attention functions
Large Language Model Transformers
E
normalcomputing
111
29
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase